2024-11-05 18:21:50,012 [ 423707 ] INFO : ClickHouse root is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse (runner:42, check_args_and_update_paths) 2024-11-05 18:21:50,012 [ 423707 ] INFO : Cases dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:86, check_args_and_update_paths) 2024-11-05 18:21:50,012 [ 423707 ] INFO : utils dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/utils (runner:97, check_args_and_update_paths) 2024-11-05 18:21:50,012 [ 423707 ] INFO : base_configs_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/programs/server, binary: /home/ubuntu/_work/_temp/test/build/clickhouse, cases_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:99, check_args_and_update_paths) clickhouse_integration_tests_volume Running pytest container as: 'docker run --rm --name clickhouse_integration_tests_b1iowx --privileged --dns-search='.' --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-odbc-bridge:/clickhouse-odbc-bridge --volume=/home/ubuntu/_work/_temp/test/build/clickhouse:/clickhouse --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-library-bridge:/clickhouse-library-bridge --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/programs/server:/clickhouse-config --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration:/ClickHouse/tests/integration --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/backupview:/ClickHouse/utils/backupview --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/grpc-client/pb2:/ClickHouse/utils/grpc-client/pb2 --volume=/run:/run/host:ro --volume=clickhouse_integration_tests_volume:/var/lib/docker -e DOCKER_DOTNET_CLIENT_TAG=11de0b29a15d -e DOCKER_HELPER_TAG=2cffe1eae894 -e DOCKER_BASE_TAG=1d1c87eee2db -e DOCKER_KERBERIZED_HADOOP_TAG=ce74919e88f5 -e DOCKER_KERBEROS_KDC_TAG=9391ecdee8d7 -e DOCKER_MYSQL_GOLANG_CLIENT_TAG=9bec2a638e6e -e DOCKER_MYSQL_JAVA_CLIENT_TAG=766bff31cfe4 -e DOCKER_MYSQL_JS_CLIENT_TAG=41ba7c2ec2a1 -e DOCKER_MYSQL_PHP_CLIENT_TAG=88be89c1e3b6 -e DOCKER_NGINX_DAV_TAG=b55ac9cd7519 -e DOCKER_POSTGRESQL_JAVA_CLIENT_TAG=a4eff5c7f4d6 -e DOCKER_PYTHON_BOTTLE_TAG=a2d3dc777d0c -e DOCKER_CLIENT_TIMEOUT=300 -e COMPOSE_HTTP_TIMEOUT=600 -e CLICKHOUSE_USE_OLD_ANALYZER=1 -e PYTHONUNBUFFERED=1 -e PYTEST_ADDOPTS=" -rfEps --run-id=2 --color=no --durations=0 test_postgresql_replica_database_engine_2/test.py::test_failed_load_from_snapshot test_postgresql_replica_database_engine_2/test.py::test_quoting_publication -vvv" altinityinfra/integration-tests-runner:9d492c2eec24 '. Start tests ============================= test session starts ============================== platform linux -- Python 3.10.12, pytest-7.4.4, pluggy-1.5.0 -- /usr/bin/python3 cachedir: .pytest_cache rootdir: /ClickHouse/tests/integration configfile: pytest.ini plugins: repeat-0.9.3, reportlog-0.4.0, random-0.2, timeout-2.2.0, order-1.0.1, xdist-3.5.0 timeout: 900.0s timeout method: signal timeout func_only: False collecting ... collected 2 items test_postgresql_replica_database_engine_2/test.py::test_failed_load_from_snapshot SKIPPEDl error) [ 50%] test_postgresql_replica_database_engine_2/test.py::test_quoting_publication FAILED [100%] =================================== FAILURES =================================== ___________________________ test_quoting_publication ___________________________ started_cluster = def test_quoting_publication(started_cluster): postgres_database = "postgres-postgres" pg_manager3 = PostgresManager() pg_manager3.init( instance, cluster.postgres_ip, cluster.postgres_port, default_database=postgres_database, ) NUM_TABLES = 5 materialized_database = "test-database" pg_manager3.create_and_fill_postgres_tables(NUM_TABLES, 10000) check_table_name_1 = "postgresql-replica-5" pg_manager3.create_and_fill_postgres_table(check_table_name_1) > pg_manager3.create_materialized_db( ip=started_cluster.postgres_ip, port=started_cluster.postgres_port, materialized_database=materialized_database, ) test_postgresql_replica_database_engine_2/test.py:1117: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ helpers/postgres_utility.py:248: in create_materialized_db self.instance.query(f"DROP DATABASE IF EXISTS {materialized_database}") helpers/cluster.py:3451: in query return self.client.query( helpers/client.py:36: in wrap return func(self, *args, **kwargs) helpers/client.py:74: in query ).get_answer() _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = def get_answer(self): self.process.wait(timeout=DEFAULT_QUERY_TIMEOUT) self.stdout_file.seek(0) self.stderr_file.seek(0) stdout = self.stdout_file.read().decode("utf-8", errors="replace") stderr = self.stderr_file.read().decode("utf-8", errors="replace") if ( self.timer is not None and not self.process_finished_before_timeout and not self.ignore_error ): logging.debug(f"Timed out. Last stdout:{stdout}, stderr:{stderr}") raise QueryTimeoutExceedException("Client timed out!") if ( self.process.returncode != 0 or self.remove_trash_from_stderr(stderr) ) and not self.ignore_error: > raise QueryRuntimeException( "Client failed! Return code: {}, stderr: {}".format( self.process.returncode, stderr ), self.process.returncode, stderr, ) E helpers.client.QueryRuntimeException: Client failed! Return code: 62, stderr: Code: 62. DB::Exception: Syntax error: failed at position 29 ('-'): -database. Expected one of: ON, NO DELAY, SYNC, INTO OUTFILE, FORMAT, SETTINGS, end of query. (SYNTAX_ERROR), Stack trace (when copying this message, always include the lines below): E E 0. ./contrib/llvm-project/libcxx/include/exception:141: Poco::Exception::Exception(String const&, int) @ 0x000000003a509d73 E 1. ./build_docker/./src/Common/Exception.cpp:96: DB::Exception::Exception(DB::Exception::MessageMasked&&, int, bool) @ 0x000000001cb535b0 E 2. ./contrib/llvm-project/libcxx/include/string:1499: DB::Exception::createDeprecated(String const&, int, bool) @ 0x000000001cc877d0 E 3. ./build_docker/./src/Parsers/parseQuery.cpp:373: DB::parseQueryAndMovePosition(DB::IParser&, char const*&, char const*, String const&, bool, unsigned long, unsigned long, unsigned long) @ 0x00000000341fe014 E 4. ./build_docker/./src/Client/ClientBase.cpp:371: DB::ClientBase::parseQuery(char const*&, char const*, bool) const @ 0x00000000323230bd E 5. ./build_docker/./src/Client/ClientBase.cpp:2087: DB::ClientBase::analyzeMultiQueryText(char const*&, char const*&, char const*, String&, std::shared_ptr&, String const&, std::unique_ptr>&) @ 0x000000003235223d E 6. ./build_docker/./src/Client/ClientBase.cpp:2176: DB::ClientBase::executeMultiQuery(String const&) @ 0x00000000323538fe E 7. ./build_docker/./src/Client/ClientBase.cpp:0: DB::ClientBase::processQueryText(String const&) @ 0x0000000032356ff4 E 8. ./contrib/llvm-project/libcxx/include/string:1499: DB::ClientBase::runNonInteractive() @ 0x0000000032361b68 E 9. ./build_docker/./programs/client/Client.cpp:0: DB::Client::main(std::vector> const&) @ 0x000000001d1567dc E 10. ./build_docker/./base/poco/Util/src/Application.cpp:0: Poco::Util::Application::run() @ 0x000000003a24c61f E 11. ./build_docker/./programs/client/Client.cpp:0: mainEntryClickHouseClient(int, char**) @ 0x000000001d1890fc E 12. ./contrib/llvm-project/libcxx/include/vector:580: main @ 0x000000000a7ffbcb E 13. ? @ 0x00007f44d8373d90 E 14. ? @ 0x00007f44d8373e40 E 15. _start @ 0x000000000a72c5ee helpers/client.py:239: QueryRuntimeException ---------------------------- Captured stdout setup ----------------------------- PostgreSQL is available - running test ----------------------------- Captured stdout call ----------------------------- Query: CREATE TABLE IF NOT EXISTS "postgresql_replica_0" ( key Integer NOT NULL, value Integer, PRIMARY KEY(key)) Query: CREATE TABLE IF NOT EXISTS "postgresql_replica_1" ( key Integer NOT NULL, value Integer, PRIMARY KEY(key)) Query: CREATE TABLE IF NOT EXISTS "postgresql_replica_2" ( key Integer NOT NULL, value Integer, PRIMARY KEY(key)) Query: CREATE TABLE IF NOT EXISTS "postgresql_replica_3" ( key Integer NOT NULL, value Integer, PRIMARY KEY(key)) Query: CREATE TABLE IF NOT EXISTS "postgresql_replica_4" ( key Integer NOT NULL, value Integer, PRIMARY KEY(key)) Query: CREATE TABLE IF NOT EXISTS "postgresql-replica-5" ( key Integer NOT NULL, value Integer, PRIMARY KEY(key)) ------------------------------ Captured log call ------------------------------- 2024-11-05 18:22:12 [ 491 ] DEBUG : Executing query DROP DATABASE IF EXISTS "postgres-postgres" on instance (cluster.py:3450, query) 2024-11-05 18:22:12 [ 491 ] DEBUG : Executing query CREATE DATABASE "postgres-postgres" ENGINE = PostgreSQL('172.16.1.2:5432', 'postgres-postgres', 'postgres', 'mysecretpassword') on instance (cluster.py:3450, query) 2024-11-05 18:22:12 [ 491 ] DEBUG : Executing query INSERT INTO `postgres-postgres`.postgresql_replica_0 SELECT number, number from numbers(10000) on instance (cluster.py:3450, query) 2024-11-05 18:22:13 [ 491 ] DEBUG : Executing query INSERT INTO `postgres-postgres`.postgresql_replica_1 SELECT number, number from numbers(10000) on instance (cluster.py:3450, query) 2024-11-05 18:22:13 [ 491 ] DEBUG : Executing query INSERT INTO `postgres-postgres`.postgresql_replica_2 SELECT number, number from numbers(10000) on instance (cluster.py:3450, query) 2024-11-05 18:22:13 [ 491 ] DEBUG : Executing query INSERT INTO `postgres-postgres`.postgresql_replica_3 SELECT number, number from numbers(10000) on instance (cluster.py:3450, query) 2024-11-05 18:22:13 [ 491 ] DEBUG : Executing query INSERT INTO `postgres-postgres`.postgresql_replica_4 SELECT number, number from numbers(10000) on instance (cluster.py:3450, query) 2024-11-05 18:22:14 [ 491 ] DEBUG : Executing query INSERT INTO `postgres-postgres`.`postgresql-replica-5` SELECT number, number from numbers(50) on instance (cluster.py:3450, query) 2024-11-05 18:22:14 [ 491 ] DEBUG : Executing query DROP DATABASE IF EXISTS test-database on instance (cluster.py:3450, query) ---------------------------- Captured log teardown ----------------------------- 2024-11-05 18:22:14 [ 491 ] DEBUG : Executing query DROP DATABASE IF EXISTS "postgres_database" on instance (cluster.py:3450, query) 2024-11-05 18:22:15 [ 491 ] DEBUG : Executing query DROP DATABASE IF EXISTS "postgres_database" on instance (cluster.py:3450, query) 2024-11-05 18:22:15 [ 491 ] DEBUG : Executing query CREATE DATABASE "postgres_database" ENGINE = PostgreSQL('172.16.1.2:5432', 'postgres_database', 'postgres', 'mysecretpassword') on instance (cluster.py:3450, query) 2024-11-05 18:22:15 [ 491 ] DEBUG : Command:['docker-compose', '--env-file', '/ClickHouse/tests/integration/test_postgresql_replica_database_engine_2/_instances_2/.env', '--project-name', 'roottestpostgresqlreplicadatabaseengine2', '--file', '/ClickHouse/tests/integration/test_postgresql_replica_database_engine_2/_instances_2/instance/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_postgres.yml', '--file', '/ClickHouse/tests/integration/test_postgresql_replica_database_engine_2/_instances_2/instance2/docker-compose.yml', 'stop', '--timeout', '20'] (cluster.py:109, run_and_check) 2024-11-05 18:22:16 [ 491 ] DEBUG : Stderr:Stopping roottestpostgresqlreplicadatabaseengine2_instance_1 ... (cluster.py:119, run_and_check) 2024-11-05 18:22:16 [ 491 ] DEBUG : Stderr:Stopping roottestpostgresqlreplicadatabaseengine2_instance2_1 ... (cluster.py:119, run_and_check) 2024-11-05 18:22:16 [ 491 ] DEBUG : Stderr:Stopping roottestpostgresqlreplicadatabaseengine2_postgres1_1 ... (cluster.py:119, run_and_check) 2024-11-05 18:22:16 [ 491 ] DEBUG : Stderr:Stopping roottestpostgresqlreplicadatabaseengine2_postgres1_1 ... done (cluster.py:119, run_and_check) 2024-11-05 18:22:16 [ 491 ] DEBUG : Stderr:Stopping roottestpostgresqlreplicadatabaseengine2_instance2_1 ... done (cluster.py:119, run_and_check) 2024-11-05 18:22:16 [ 491 ] DEBUG : Stderr:Stopping roottestpostgresqlreplicadatabaseengine2_instance_1 ... done (cluster.py:119, run_and_check) 2024-11-05 18:22:16 [ 491 ] DEBUG : Command:['bash', '-c', '[ -f /ClickHouse/tests/integration/test_postgresql_replica_database_engine_2/_instances_2/instance/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_postgresql_replica_database_engine_2/_instances_2/instance/logs/stderr.log* || true'] (cluster.py:109, run_and_check) 2024-11-05 18:22:16 [ 491 ] DEBUG : Command:['bash', '-c', '[ -f /ClickHouse/tests/integration/test_postgresql_replica_database_engine_2/_instances_2/instance2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_postgresql_replica_database_engine_2/_instances_2/instance2/logs/stderr.log* || true'] (cluster.py:109, run_and_check) 2024-11-05 18:22:16 [ 491 ] DEBUG : Command:['docker-compose', '--env-file', '/ClickHouse/tests/integration/test_postgresql_replica_database_engine_2/_instances_2/.env', '--project-name', 'roottestpostgresqlreplicadatabaseengine2', '--file', '/ClickHouse/tests/integration/test_postgresql_replica_database_engine_2/_instances_2/instance/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_postgres.yml', '--file', '/ClickHouse/tests/integration/test_postgresql_replica_database_engine_2/_instances_2/instance2/docker-compose.yml', 'down', '--volumes'] (cluster.py:109, run_and_check) 2024-11-05 18:22:17 [ 491 ] DEBUG : Stderr:Removing roottestpostgresqlreplicadatabaseengine2_instance_1 ... (cluster.py:119, run_and_check) 2024-11-05 18:22:17 [ 491 ] DEBUG : Stderr:Removing roottestpostgresqlreplicadatabaseengine2_instance2_1 ... (cluster.py:119, run_and_check) 2024-11-05 18:22:17 [ 491 ] DEBUG : Stderr:Removing roottestpostgresqlreplicadatabaseengine2_postgres1_1 ... (cluster.py:119, run_and_check) 2024-11-05 18:22:17 [ 491 ] DEBUG : Stderr:Removing roottestpostgresqlreplicadatabaseengine2_instance2_1 ... done (cluster.py:119, run_and_check) 2024-11-05 18:22:17 [ 491 ] DEBUG : Stderr:Removing roottestpostgresqlreplicadatabaseengine2_instance_1 ... done (cluster.py:119, run_and_check) 2024-11-05 18:22:17 [ 491 ] DEBUG : Stderr:Removing roottestpostgresqlreplicadatabaseengine2_postgres1_1 ... done (cluster.py:119, run_and_check) 2024-11-05 18:22:17 [ 491 ] DEBUG : Stderr:Removing network roottestpostgresqlreplicadatabaseengine2_default (cluster.py:119, run_and_check) 2024-11-05 18:22:17 [ 491 ] DEBUG : Cleanup called (cluster.py:797, cleanup) 2024-11-05 18:22:17 [ 491 ] DEBUG : Docker networks for project roottestpostgresqlreplicadatabaseengine2 are NETWORK ID NAME DRIVER SCOPE (cluster.py:776, print_all_docker_pieces) 2024-11-05 18:22:17 [ 491 ] DEBUG : Docker containers for project roottestpostgresqlreplicadatabaseengine2 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:784, print_all_docker_pieces) 2024-11-05 18:22:17 [ 491 ] DEBUG : Docker volumes for project roottestpostgresqlreplicadatabaseengine2 are DRIVER VOLUME NAME (cluster.py:792, print_all_docker_pieces) 2024-11-05 18:22:17 [ 491 ] DEBUG : Command:docker container list --all --filter name='^/roottestpostgresqlreplicadatabaseengine2_.*_1$' --format '{{.ID}}:{{.Names}}' (cluster.py:109, run_and_check) 2024-11-05 18:22:17 [ 491 ] DEBUG : Unstopped containers: {} (cluster.py:811, cleanup) 2024-11-05 18:22:17 [ 491 ] DEBUG : No running containers for project: roottestpostgresqlreplicadatabaseengine2 (cluster.py:825, cleanup) 2024-11-05 18:22:17 [ 491 ] DEBUG : Trying to prune unused networks... (cluster.py:831, cleanup) 2024-11-05 18:22:17 [ 491 ] DEBUG : Trying to prune unused images... (cluster.py:847, cleanup) 2024-11-05 18:22:17 [ 491 ] DEBUG : Command:['docker', 'image', 'prune', '-f'] (cluster.py:109, run_and_check) 2024-11-05 18:22:17 [ 491 ] DEBUG : Stdout:Total reclaimed space: 0B (cluster.py:117, run_and_check) 2024-11-05 18:22:17 [ 491 ] DEBUG : Images pruned (cluster.py:850, cleanup) 2024-11-05 18:22:17 [ 491 ] DEBUG : Trying to prune unused volumes... (cluster.py:856, cleanup) 2024-11-05 18:22:17 [ 491 ] DEBUG : Command:['docker volume ls | wc -l'] (cluster.py:109, run_and_check) 2024-11-05 18:22:17 [ 491 ] DEBUG : Stdout:1 (cluster.py:117, run_and_check) ============================== slowest durations =============================== 18.83s setup test_postgresql_replica_database_engine_2/test.py::test_failed_load_from_snapshot 2.70s teardown test_postgresql_replica_database_engine_2/test.py::test_quoting_publication 2.31s call test_postgresql_replica_database_engine_2/test.py::test_quoting_publication 0.72s teardown test_postgresql_replica_database_engine_2/test.py::test_failed_load_from_snapshot 0.17s call test_postgresql_replica_database_engine_2/test.py::test_failed_load_from_snapshot 0.00s setup test_postgresql_replica_database_engine_2/test.py::test_quoting_publication =========================== short test summary info ============================ FAILED test_postgresql_replica_database_engine_2/test.py::test_quoting_publication SKIPPED [1] test_postgresql_replica_database_engine_2/test.py:900: Sanitizers and debug mode are skipped, because this test thrown logical error ======================== 1 failed, 1 skipped in 24.88s ========================= Traceback (most recent call last): File "/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration/./runner", line 437, in subprocess.check_call(cmd, shell=True) File "/usr/lib/python3.10/subprocess.py", line 369, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command 'docker run --rm --name clickhouse_integration_tests_b1iowx --privileged --dns-search='.' --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-odbc-bridge:/clickhouse-odbc-bridge --volume=/home/ubuntu/_work/_temp/test/build/clickhouse:/clickhouse --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-library-bridge:/clickhouse-library-bridge --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/programs/server:/clickhouse-config --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration:/ClickHouse/tests/integration --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/backupview:/ClickHouse/utils/backupview --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/grpc-client/pb2:/ClickHouse/utils/grpc-client/pb2 --volume=/run:/run/host:ro --volume=clickhouse_integration_tests_volume:/var/lib/docker -e DOCKER_DOTNET_CLIENT_TAG=11de0b29a15d -e DOCKER_HELPER_TAG=2cffe1eae894 -e DOCKER_BASE_TAG=1d1c87eee2db -e DOCKER_KERBERIZED_HADOOP_TAG=ce74919e88f5 -e DOCKER_KERBEROS_KDC_TAG=9391ecdee8d7 -e DOCKER_MYSQL_GOLANG_CLIENT_TAG=9bec2a638e6e -e DOCKER_MYSQL_JAVA_CLIENT_TAG=766bff31cfe4 -e DOCKER_MYSQL_JS_CLIENT_TAG=41ba7c2ec2a1 -e DOCKER_MYSQL_PHP_CLIENT_TAG=88be89c1e3b6 -e DOCKER_NGINX_DAV_TAG=b55ac9cd7519 -e DOCKER_POSTGRESQL_JAVA_CLIENT_TAG=a4eff5c7f4d6 -e DOCKER_PYTHON_BOTTLE_TAG=a2d3dc777d0c -e DOCKER_CLIENT_TIMEOUT=300 -e COMPOSE_HTTP_TIMEOUT=600 -e CLICKHOUSE_USE_OLD_ANALYZER=1 -e PYTHONUNBUFFERED=1 -e PYTEST_ADDOPTS=" -rfEps --run-id=2 --color=no --durations=0 test_postgresql_replica_database_engine_2/test.py::test_failed_load_from_snapshot test_postgresql_replica_database_engine_2/test.py::test_quoting_publication -vvv" altinityinfra/integration-tests-runner:9d492c2eec24 ' returned non-zero exit status 1.